Supplement to: Scaling the Indian Buffet Process via Submodular Maximization

نویسنده

  • Zoubin Ghahramani
چکیده

Here we discuss the “shifted” equivalence class of binary matrices first proposed by Ding et al. (2010). For a given N ×K binary matrix Z, the equivalence class for this binary matrix [Z] is obtained by shifting allzero columns to the right of the non-zero columns while maintaining the non-zero column orderings, see Figure 1. Placing independent Beta( α K , 1) priors on the Bernoulli entries of Z and integrating over these priors yields the following probability for Z, see Eq. 27 in Griffiths & Ghahramani (2005):

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Scaling the Indian Buffet Process via Submodular Maximization

Inference for latent feature models is inherently difficult as the inference space grows exponentially with the size of the input data and number of latent features. In this work, we use Kurihara & Welling (2008)’s maximization-expectation framework to perform approximate MAP inference for linearGaussian latent feature models with an Indian Buffet Process (IBP) prior. This formulation yields a ...

متن کامل

Scaling Submodular Maximization via Pruned Submodularity Graphs

We propose a new randomized pruning method (called “submodular sparsification (SS)”) to reduce the cost of submodular maximization. The pruning is applied via a “submodularity graph” over the n ground elements, where each directed edge is associated with a pairwise dependency defined by the submodular function. In each step, SS prunes a 1 1/ p c (for c > 1) fraction of the nodes using weights o...

متن کامل

Multi-document Summarization via Budgeted Maximization of Submodular Functions

We treat the text summarization problem as maximizing a submodular function under a budget constraint. We show, both theoretically and empirically, a modified greedy algorithm can efficiently solve the budgeted submodular maximization problem near-optimally, and we derive new approximation bounds in doing so. Experiments on DUC’04 task show that our approach is superior to the bestperforming me...

متن کامل

Learning Sparse Combinatorial Representations via Two-stage Submodular Maximization

We consider the problem of learning sparse representations of data sets, where the goal is to reduce a data set in manner that optimizes multiple objectives. Motivated by applications of data summarization, we develop a new model which we refer to as the two-stage submodular maximization problem. This task can be viewed as a combinatorial analogue of representation learning problems such as dic...

متن کامل

Probabilistic Submodular Maximization in Sub-Linear Time

In this paper, we consider optimizing submodular functions that are drawn from some unknown distribution. This setting arises, e.g., in recommender systems, where the utility of a subset of items may depend on a user-specific submodular utility function. In modern applications, the ground set of items is often so large that even the widely used (lazy) greedy algorithm is not efficient enough. A...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

عنوان ژورنال:

دوره   شماره 

صفحات  -

تاریخ انتشار 2013